- Home
- Search Results
- Page 1 of 1
Search for: All records
-
Total Resources1
- Resource Type
-
0000000001000000
- More
- Availability
-
01
- Author / Contributor
- Filter by Author / Creator
-
-
Blake, Cayden (1)
-
Ford, Chase (1)
-
Manner, Benson (1)
-
Webb, Benjamin (1)
-
#Tyler Phillips, Kenneth E. (0)
-
#Willis, Ciara (0)
-
& Abreu-Ramos, E. D. (0)
-
& Abramson, C. I. (0)
-
& Abreu-Ramos, E. D. (0)
-
& Adams, S.G. (0)
-
& Ahmed, K. (0)
-
& Ahmed, Khadija. (0)
-
& Aina, D.K. Jr. (0)
-
& Akcil-Okan, O. (0)
-
& Akuom, D. (0)
-
& Aleven, V. (0)
-
& Andrews-Larson, C. (0)
-
& Archibald, J. (0)
-
& Arnett, N. (0)
-
& Arya, G. (0)
-
- Filter by Editor
-
-
& Spizer, S. M. (0)
-
& . Spizer, S. (0)
-
& Ahn, J. (0)
-
& Bateiha, S. (0)
-
& Bosch, N. (0)
-
& Brennan K. (0)
-
& Brennan, K. (0)
-
& Chen, B. (0)
-
& Chen, Bodong (0)
-
& Drown, S. (0)
-
& Ferretti, F. (0)
-
& Higgins, A. (0)
-
& J. Peters (0)
-
& Kali, Y. (0)
-
& Ruiz-Arias, P.M. (0)
-
& S. Spitzer (0)
-
& Sahin. I. (0)
-
& Spitzer, S. (0)
-
& Spitzer, S.M. (0)
-
(submitted - in Review for IEEE ICASSP-2024) (0)
-
-
Have feedback or suggestions for a way to improve these results?
!
Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
-
Abstract One of the more popular optimization methods in current use is the Adam optimizer. This is due, at least in part, to its effectiveness as a training algorithm for deep neural networks, which are associated with many machine learning tasks. In this paper, we introduce time delays into the Adam optimizer. Time delays typically have an adverse effect on dynamical systems, including optimizers, slowing the system’s rate of convergence and potentially causing instabilities. However, our numerical experiments indicate that introducing time-delays into the Adam optimizer can significantly improve its performance, resulting in an often much smaller loss-value. Perhaps more surprising is that this improvement often scales with dimension-the higher the dimension the greater the advantage of using time delays in improving loss-values. Along with describing these results we show that, for the time-delays we consider, the temporal complexity of the delayed Adam optimizer remains the same as the undelayed optimizer and that the algorithm’s spatial complexity scales linearly in the length of the largest time-delay. Last, we extend the theory of intrinsic stability to give a criterion under which the minima, either local or global, associated with the delayed Adam optimizer are stable.more » « lessFree, publicly-accessible full text available July 10, 2026
An official website of the United States government
